Improving Domain-Generalized Few-Shot Text Classification with Multi-Level Distributional Signatures

نویسندگان

چکیده

Domain-generalized few-shot text classification (DG-FSTC) is a new setting for (FSTC). In DG-FSTC, the model meta-trained on multi-domain dataset, and meta-tested unseen datasets with different domains. However, previous methods mostly construct semantic representations by learning from words directly, which limited in domain adaptability. this study, we enhance adaptability of utilizing distributional signatures texts that indicate domain-related features specific We propose Multi-level Distributional Signatures based model, namely MultiDS. Firstly, inspired pretrained language models, compute an extra large news corpus, denote these as domain-agnostic features. Then calculate same class, respectively. These two kinds information are regarded domain-specific class-specific features, After that, fuse translate three into word-level attention values, enables to capture informative changes. addition, utilize calibration feature The vectors produced word embeddings help adapt various Extensive experiments performed four benchmarks. results demonstrate our proposed method beats state-of-the-art average improvement 1.41% datasets. Compared five competitive baselines, achieves best performance. ablation studies prove effectiveness each module.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Few-Shot Adversarial Domain Adaptation

This work provides a framework for addressing the problem of supervised domain adaptation with deep models. The main idea is to exploit adversarial learning to learn an embedded subspace that simultaneously maximizes the confusion between two domains while semantically aligning their embedding. The supervised setting becomes attractive especially when there are only a few target data samples th...

متن کامل

Few-Shot Text Classification with Pre-Trained Word Embeddings and a Human in the Loop

Most of the literature around text classification treats it as a supervised learning problem: given a corpus of labeled documents, train a classifier such that it can accurately predict the classes of unseen documents. In industry, however, it is not uncommon for a business to have entire corpora of documents where few or none have been classified, or where existing classifications have become ...

متن کامل

Improving Multi-Document Summarization via Text Classification

Developed so far, multi-document summarization has reached its bottleneck due to the lack of sufficient training data and diverse categories of documents. Text classification just makes up for these deficiencies. In this paper, we propose a novel summarization system called TCSum, which leverages plentiful text classification data to improve the performance of multi-document summarization. TCSu...

متن کامل

Few-shot Classification by Learning Disentangled Representations

Machine learning has improved state-of-the art performance in numerous domains, by using large amounts of data. In reality, labelled data is often not available for the task of interest. A fundamental problem of artificial intelligence is finding a representation that can generalize to never seen before classes. In this research, the power of generative models is combined with disentangled repr...

متن کامل

Improving Multi-class Text Classification with Naive Bayes

There are numerous text documents available in electronic form. More and more are becoming available every day. Such documents represent a massive amount of information that is easily accessible. Seeking value in this huge collection requires organization; much of the work of organizing documents can be automated through text classification. The accuracy and our understanding of such systems gr...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Applied sciences

سال: 2023

ISSN: ['2076-3417']

DOI: https://doi.org/10.3390/app13021202